GA Guard Lite is a lightweight open-source moderation model developed by General Analysis, specifically designed to detect prohibited content in the outputs of language models. This model can identify seven categories of security risks, including illegal activities, hate speech, and personal information leakage, providing efficient content security protection for AI applications.
Natural Language Processing
TransformersEnglish